PicSOM Experiments in TRECVID 2009

نویسندگان

  • Mats Sjöberg
  • Ville Viitaniemi
  • Markus Koskela
  • Jorma Laaksonen
چکیده

Our experiments in TRECVID 2009 include participation in the high-level feature extraction and automatic search tasks. In the high-level feature extraction task, we used a feature fusion-based general system architecture utilizing a large number of SVM detectors, followed by a post-processing stage utilizing the concepts’ temporal and inter-concept co-occurrences. We submitted the following six runs: • PicSOM.base: Baseline run using our SOM-based HLF detection method • PicSOM.A-ngram: Baseline SVM-based run using HLF-wise geometric mean fusion and temporal n-gram post-processing • PicSOM.B-ngram: As previous, but includes also early fusion, multi-fold SFBS fusion, and more elaborate SVM training • PicSOM.E-ngram: As previous, but includes two-stage fusion utilizing cross-concept co-occurrence • PicSOM.spec-ngram: A run where the used method was selected for each HLF separately using cross-validation • PicSOM.spec-any: As previous, but the post-processing used also clustering-based inter-concept co-occurrence analysis The results show that feature fusion can consistently outperform all single features, multi-fold SFBS performed best of the tested fusion methods, and that temporal n-gram analysis is beneficial. Early fusion, and post-processing based on inter-concept co-occurrences did not improve the performance. In the search task, we concentrated on the fully-automatic runs and standard search task. We combined ASR/MT text search and concept-based retrieval. If none of the concept models could be matched with the query, we used content-based retrieval based on the video and image examples instead. We submitted the following ten fully-automatic runs: • F_A_N_PicSOM_1_10: text search baseline • F_A_N_PicSOM_2_9: visual baseline • F_A_N_PicSOM_3_8: own concepts • F_A_N_PicSOM_4_7: own concepts + text search • F_A_N_PicSOM_5_6: donated concepts • F_A_N_PicSOM_6_5: donated concepts + text search • F_A_N_PicSOM_7_4: own + donated concepts • F_A_N_PicSOM_8_3: own + donated concepts + text search • F_A_N_PicSOM_9_2: own + donated (dupl.) concepts • F_A_N_PicSOM_10_1: own + donated (dupl.) concepts + text search In the above list, “own” concepts refer to our own HLF detectors and “donated” concepts consist of MediaMill (MM) concepts + CU-VIREO374 concepts. In other than the last two runs, CU-VIREO374 are only used for words for which no MediaMill concept could be matched. The results show again that concept-based retrieval performed better than content-based search alone. Text search made a small improvement in combination with other modalities, but performed really badly on its own. Concept-selection was done both with word-matching and example-based matching, i.e. selecting concepts based on how well they would fit our own concept models.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Concept-based Video Search with the Picsom Multi- Media Retrieval System

In this report we describe the structure of the PicSOM multimedia retrieval system and elaborate on its automatic concept detection and video search subsystems. We evaluate several alternative techniques for implementing these two components of the PicSOM system in a comprehensive series of experiments employing the large-scale setups of the TRECVID video retrieval evaluation campaigns of 2008 ...

متن کامل

PicSOM Experiments in TRECVID 2014

Our experiments in TRECVID 2014 include successful participation in the Semantic Indexing (SIN) task and unsuccessful participation in the Multimedia Event Detection (MED) and Multimedia Event Recounting (MER) tasks. In semantic indexing, we participated in the main task only. We extended our last year’s set of features with SIFT descriptors encoded with Fisher vectors and VLAD, and a total of ...

متن کامل

PicSOM Experiments in TRECVID 2006

Our experiments in TRECVID 2006 include participation in the shot boundary detection, high-level feature extraction, and search tasks, using a common system framework based on multiple parallel Self-Organizing Maps (SOMs). In the shot boundary detection task we projected feature vectors calculated from successive frames on parallel SOMs and monitored the trajectories to detect the shot boundari...

متن کامل

REGIMVID at TRECVID 2009: Semantic Access to Multimedia Data

In this paper we describe our TRECVID 2009 video retrieval experiments. The REGIMVID team participated in two tasks: High Level Feature Extraction and Automatic Search. Our TRECVID 2009 experiments focus on increasing the robustness of a small set of sensors and the relevance of the results using a probabilistic weighting of learning examples.

متن کامل

Using Image Segments in PicSOM CBIR System

The content-based image retrieval (CBIR) system PicSOM uses a variety of low-level visual features as an indexing mechanism for an image database. In this paper we describe the implementation of segmentation into the PicSOM framework. That is, we have modified the system to use image segments as a supplement to entire images in order to improve the retrieval accuracy. In a series of experiments...

متن کامل

PicSOM Experiments in TRECVID 2010

Our experiments in TRECVID 2010 include participation in the semantic indexing and known-item search tasks. In the semantic indexing task we implemented SVM-based classifiers on five different low-level visual features extracted from the keyframes. In addition to the main keyframes provided by NIST, we also extracted and analysed additional frames from longer shots. The feature-wise classifiers...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2009